40 research outputs found
Fundamental activity constraints lead to specific interpretations of the connectome
The continuous integration of experimental data into coherent models of the
brain is an increasing challenge of modern neuroscience. Such models provide a
bridge between structure and activity, and identify the mechanisms giving rise
to experimental observations. Nevertheless, structurally realistic network
models of spiking neurons are necessarily underconstrained even if experimental
data on brain connectivity are incorporated to the best of our knowledge.
Guided by physiological observations, any model must therefore explore the
parameter ranges within the uncertainty of the data. Based on simulation
results alone, however, the mechanisms underlying stable and physiologically
realistic activity often remain obscure. We here employ a mean-field reduction
of the dynamics, which allows us to include activity constraints into the
process of model construction. We shape the phase space of a multi-scale
network model of the vision-related areas of macaque cortex by systematically
refining its connectivity. Fundamental constraints on the activity, i.e.,
prohibiting quiescence and requiring global stability, prove sufficient to
obtain realistic layer- and area-specific activity. Only small adaptations of
the structure are required, showing that the network operates close to an
instability. The procedure identifies components of the network critical to its
collective dynamics and creates hypotheses for structural data and future
experiments. The method can be applied to networks involving any neuron model
with a known gain function.Comment: J. Schuecker and M. Schmidt contributed equally to this wor
Integration of continuous-time dynamics in a spiking neural network simulator
Contemporary modeling approaches to the dynamics of neural networks consider
two main classes of models: biologically grounded spiking neurons and
functionally inspired rate-based units. The unified simulation framework
presented here supports the combination of the two for multi-scale modeling
approaches, the quantitative validation of mean-field approaches by spiking
network simulations, and an increase in reliability by usage of the same
simulation code and the same network model specifications for both model
classes. While most efficient spiking simulations rely on the communication of
discrete events, rate models require time-continuous interactions between
neurons. Exploiting the conceptual similarity to the inclusion of gap junctions
in spiking network simulations, we arrive at a reference implementation of
instantaneous and delayed interactions between rate-based models in a spiking
network simulator. The separation of rate dynamics from the general connection
and communication infrastructure ensures flexibility of the framework. We
further demonstrate the broad applicability of the framework by considering
various examples from the literature ranging from random networks to neural
field models. The study provides the prerequisite for interactions between
rate-based and spiking models in a joint simulation
Conditions for wave trains in spiking neural networks
Spatiotemporal patterns such as traveling waves are frequently observed in
recordings of neural activity. The mechanisms underlying the generation of such
patterns are largely unknown. Previous studies have investigated the existence
and uniqueness of different types of waves or bumps of activity using
neural-field models, phenomenological coarse-grained descriptions of
neural-network dynamics. But it remains unclear how these insights can be
transferred to more biologically realistic networks of spiking neurons, where
individual neurons fire irregularly. Here, we employ mean-field theory to
reduce a microscopic model of leaky integrate-and-fire (LIF) neurons with
distance-dependent connectivity to an effective neural-field model. In contrast
to existing phenomenological descriptions, the dynamics in this neural-field
model depends on the mean and the variance in the synaptic input, both
determining the amplitude and the temporal structure of the resulting effective
coupling kernel. For the neural-field model we employ liner stability analysis
to derive conditions for the existence of spatial and temporal oscillations and
wave trains, that is, temporally and spatially periodic traveling waves. We
first prove that wave trains cannot occur in a single homogeneous population of
neurons, irrespective of the form of distance dependence of the connection
probability. Compatible with the architecture of cortical neural networks, wave
trains emerge in two-population networks of excitatory and inhibitory neurons
as a combination of delay-induced temporal oscillations and spatial
oscillations due to distance-dependent connectivity profiles. Finally, we
demonstrate quantitative agreement between predictions of the analytically
tractable neural-field model and numerical simulations of both networks of
nonlinear rate-based units and networks of LIF neurons.Comment: 36 pages, 8 figures, 4 table
Integration of Continuous-Time Dynamics in a Spiking Neural Network Simulator
Contemporary modeling approaches to the dynamics of neural networks include two important classes of models: biologically grounded spiking neuron models and functionally inspired rate-based units. We present a unified simulation framework that supports the combination of the two for multi-scale modeling, enables the quantitative validation of mean-field approaches by spiking network simulations, and provides an increase in reliability by usage of the same simulation code and the same network model specifications for both model classes. While most spiking simulations rely on the communication of discrete events, rate models require time-continuous interactions between neurons. Exploiting the conceptual similarity to the inclusion of gap junctions in spiking network simulations, we arrive at a reference implementation of instantaneous and delayed interactions between rate-based models in a spiking network simulator. The separation of rate dynamics from the general connection and communication infrastructure ensures flexibility of the framework. In addition to the standard implementation we present an iterative approach based on waveform-relaxation techniques to reduce communication and increase performance for large-scale simulations of rate-based models with instantaneous interactions. Finally we demonstrate the broad applicability of the framework by considering various examples from the literature, ranging from random networks to neural-field models. The study provides the prerequisite for interactions between rate-based and spiking models in a joint simulation